Gradient Boosting With Piece-Wise Linear Regression Trees
نویسندگان
چکیده
Gradient boosting using decision trees as base learners, so called Gradient Boosted Decision Trees (GBDT), is a very successful ensemble learning algorithm widely used across a variety of applications. Recently, various GDBT construction algorithms and implementation have been designed and heavily optimized in some very popular open sourced toolkits such as XGBoost and LightGBM. In this paper, we show that both the accuracy and efficiency of GBDT can be further enhanced by using more complex base learners. Specifically, we extend gradient boosting to use piecewise linear regression trees (PL Trees), instead of piecewise constant regression trees. We show PL Trees can accelerate convergence of GBDT. Moreover, our new algorithm fits better to modern computer architectures with powerful Single Instruction Multiple Data (SIMD) parallelism. We propose optimization techniques to speedup our algorithm. The experimental results show that GBDT with PL Trees can provide very competitive testing accuracy with comparable or less training time. Our algorithm also produces much concise tree ensembles, thus can often reduce testing time costs.
منابع مشابه
Effectiveness of ensemble machine learning over the conventional multivariable linear regression models
This paper demonstrates the effectiveness of ensemble machine learning algorithms over the conventional multivariable linear regression models including Ordinary Least Squares, Robust Linear Model, and Lasso Model. The ensemble machine learning algorithms include Adaboost, Random-Forest, Bagging, Extremely Randomized Trees, Gradient Boosting, and Extra Trees Regressor. With the progress of open...
متن کاملAn experimental evaluation of boosting methods for classification.
OBJECTIVES In clinical medicine, the accuracy achieved by classification rules is often not sufficient to justify their use in daily practice. In order to improve classifiers it has become popular to combine single classification rules into a classification ensemble. Two popular boosting methods will be compared with classical statistical approaches. METHODS Using data from a clinical study o...
متن کاملStratified Gradient Boosting for Fast Training of CRFs
Boosting has recently been shown to be a promising approach for training conditional random fields (CRFs) as it allows to efficiently induce conjunctive (even relational) features. The potentials are represented as weighted sums of regression trees that are induced using gradient tree boosting. Its large scale application, however, suffers from two drawbacks: induced trees can spoil previous ma...
متن کاملFacial Landmark Detection using Ensemble of Cascaded Regressions
This paper presents an ensemble of regressions approach for estimation of the positions of facial landmarks in frontal and near-frontal face images. Our approach learns three different cascades of regressors and fuses their predictions into one final precise estimate using Gradient Boosting Regression Trees (GBRT). The cascaded model starts from an approximate estimate of the landmark positions...
متن کاملModel-based boosting in high dimensions
SUMMARY The R add-on package mboost implements functional gradient descent algorithms (boosting) for optimizing general loss functions utilizing componentwise least squares, either of parametric linear form or smoothing splines, or regression trees as base learners for fitting generalized linear, additive and interaction models to potentially high-dimensional data. AVAILABILITY Package mboost...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1802.05640 شماره
صفحات -
تاریخ انتشار 2018